driving system
RoboDriveVLM: A Novel Benchmark and Baseline towards Robust Vision-Language Models for Autonomous Driving
Liao, Dacheng, Qi, Mengshi, Shu, Peng, Zhang, Zhining, Lin, Yuxin, Liu, Liang, Ma, Huadong
Current Vision-Language Model (VLM)-based end-to-end autonomous driving systems often leverage large language models to generate driving decisions directly based on their understanding of the current scene. However, such systems introduce multiple risks in real-world driving scenarios. T o evaluate whether VLMs are truly viable for autonomous driving, we introduce RoboDriveBench, the first robustness benchmark focused on end-to-end trajectory prediction tasks. This benchmark systematically evaluates two critical categories of real-world challenges for VLM-based end-to-end autonomous driving systems through 11 simulated scenarios encompassing various corruption types, including 6 scenarios of sensor corruption caused by environmental variations, along with 5 cases of prompt corruption resulting from human intervention and data transmission failures. Each corruption type includes 250 unique driving scenarios and 5,689 frames, resulting in 64,559 total trajectory prediction cases per evaluation. T o overcome these real-world challenges, we propose a novel VLM-based autonomous driving framework called Robo-DriveVLM, which enhances robustness by mapping more multimodal data--e.g., lidar and radar--into a unified latent space. Furthermore, we introduce a new T est-Time Adaptation (TTA) method based on cross-modal knowledge distillation to improve the robustness of VLM-based autonomous driving systems. Through extensive experiments, our work highlights the limitations of current VLM-based end-to-end autonomous driving systems and provides a more reliable solution for real-world deployment. Source code and datasets will be released.
- Asia > Middle East > Republic of Türkiye > Karaman Province > Karaman (0.04)
- Asia > China > Beijing > Beijing (0.04)
- Transportation > Ground > Road (1.00)
- Information Technology > Robotics & Automation (1.00)
- Automobiles & Trucks (1.00)
Integration of Computer Vision with Adaptive Control for Autonomous Driving Using ADORE
Ahammed, Abu Shad, Hossain, Md Shahi Amran, Mukherjee, Sayeri, Obermaisser, Roman, Rahman, Md. Ziaur
Ensuring safety in autonomous driving requires a seamless integration of perception and decision making under uncertain conditions. Although computer vision (CV) models such as YOLO achieve high accuracy in detecting traffic signs and obstacles, their performance degrades in drift scenarios caused by weather variations or unseen objects. This work presents a simulated autonomous driving system that combines a context aware CV model with adaptive control using the ADORE framework. The CARLA simulator was integrated with ADORE via the ROS bridge, allowing real-time communication between perception, decision, and control modules. A simulated test case was designed in both clear and drift weather conditions to demonstrate the robust detection performance of the perception model while ADORE successfully adapted vehicle behavior to speed limits and obstacles with low response latency. The findings highlight the potential of coupling deep learning-based perception with rule-based adaptive decision making to improve automotive safety critical system.
- Europe > Germany > North Rhine-Westphalia > Arnsberg Region > Siegen (0.05)
- Asia > Middle East > Yemen > Amran Governorate > Amran (0.05)
- Asia > Bangladesh > Dhaka Division > Dhaka District > Dhaka (0.04)
- Transportation > Ground > Road (1.00)
- Information Technology > Robotics & Automation (1.00)
- Automobiles & Trucks (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.55)
HE-Drive: Human-Like End-to-End Driving with Vision Language Models
Wang, Junming, Zhang, Xingyu, Xing, Zebin, Gu, Songen, Guo, Xiaoyang, Hu, Yang, Song, Ziying, Zhang, Qian, Long, Xiaoxiao, Yin, Wei
In this paper, we propose HE-Drive: the first human-like-centric end-to-end autonomous driving system to generate trajectories that are both temporally consistent and comfortable. Recent studies have shown that imitation learning-based planners and learning-based trajectory scorers can effectively generate and select accuracy trajectories that closely mimic expert demonstrations. However, such trajectory planners and scorers face the dilemma of generating temporally inconsistent and uncomfortable trajectories. To solve the above problems, Our HE-Drive first extracts key 3D spatial representations through sparse perception, which then serves as conditional inputs for a Conditional Denoising Diffusion Probabilistic Models (DDPMs)-based motion planner to generate temporal consistency multi-modal trajectories. A Vision-Language Models (VLMs)-guided trajectory scorer subsequently selects the most comfortable trajectory from these candidates to control the vehicle, ensuring human-like end-to-end driving. Experiments show that HE-Drive not only achieves state-of-the-art performance (i.e., reduces the average collision rate by 71% than VAD) and efficiency (i.e., 1.9X faster than SparseDrive) on the challenging nuScenes and OpenScene datasets but also provides the most comfortable driving experience on real-world data.For more information, visit the project website: https://jmwang0117.github.io/HE-Drive/.
- Automobiles & Trucks (0.92)
- Transportation > Ground > Road (0.52)
Edge-Cloud Collaborative Motion Planning for Autonomous Driving with Large Language Models
Chen, Jiao, Dai, Suyan, Chen, Fangfang, Lv, Zuohong, Tang, Jianhua
Integrating large language models (LLMs) into autonomous driving enhances personalization and adaptability in open-world scenarios. However, traditional edge computing models still face significant challenges in processing complex driving data, particularly regarding real-time performance and system efficiency. To address these challenges, this study introduces EC-Drive, a novel edge-cloud collaborative autonomous driving system with data drift detection capabilities. EC-Drive utilizes drift detection algorithms to selectively upload critical data, including new obstacles and traffic pattern changes, to the cloud for processing by GPT-4, while routine data is efficiently managed by smaller LLMs on edge devices. This approach not only reduces inference latency but also improves system efficiency by optimizing communication resource use. Experimental validation confirms the system's robust processing capabilities and practical applicability in real-world driving conditions, demonstrating the effectiveness of this edge-cloud collaboration framework. Our data and system demonstration will be released at https://sites.google.com/view/ec-drive.
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- Asia > China > Guangdong Province > Guangzhou (0.04)
- Transportation > Ground > Road (1.00)
- Information Technology > Robotics & Automation (0.85)
Automated Lane Change Behavior Prediction and Environmental Perception Based on SLAM Technology
Lei, Han, Wang, Baoming, Shui, Zuwei, Yang, Peiyuan, Liang, Penghao
In addition to environmental perception sensors such as cameras, radars, etc. in the automatic driving system, the external environment of the vehicle is perceived, in fact, there is also a perception sensor that has been silently dedicated in the system, that is, the positioning module. This paper explores the application of SLAM (Simultaneous Localization and Mapping) technology in the context of automatic lane change behavior prediction and environment perception for autonomous vehicles. It discusses the limitations of traditional positioning methods, introduces SLAM technology, and compares LIDAR SLAM with visual SLAM. Real-world examples from companies like Tesla, Waymo, and Mobileye showcase the integration of AI-driven technologies, sensor fusion, and SLAM in autonomous driving systems. The paper then delves into the specifics of SLAM algorithms, sensor technologies, and the importance of automatic lane changes in driving safety and efficiency. It highlights Tesla's recent update to its Autopilot system, which incorporates automatic lane change functionality using SLAM technology. The paper concludes by emphasizing the crucial role of SLAM in enabling accurate environment perception, positioning, and decision-making for autonomous vehicles, ultimately enhancing safety and driving experience.
- North America > United States > Illinois > Champaign County > Urbana (0.14)
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- North America > United States > Illinois > Champaign County > Champaign (0.04)
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks (1.00)
4C: A Computation, Communication, and Control Co-Design Framework for CAVs
Liu, Liangkai, Liu, Shaoshan, Shi, Weisong
Connected and autonomous vehicles (CAVs) are promising due to their potential safety and efficiency benefits and have attracted massive investment and interest from government agencies, industry, and academia. With more computing and communication resources are available, both vehicles and edge servers are equipped with a set of camera-based vision sensors, also known as Visual IoT (V-IoT) techniques, for sensing and perception. Tremendous efforts have been made for achieving programmable communication, computation, and control. However, they are conducted mainly in the silo mode, limiting the responsiveness and efficiency of handling challenging scenarios in the real world. To improve the end-to-end performance, we envision that future CAVs require the co-design of communication, computation, and control. This paper presents our vision of the end-to-end design principle for CAVs, called 4C, which extends the V-IoT system by providing a unified communication, computation, and control co-design framework. With programmable communications, fine-grained heterogeneous computation, and efficient vehicle controls in 4C, CAVs can handle critical scenarios and achieve energy-efficient autonomous driving. Finally, we present several challenges to achieving the vision of the 4C framework.
- North America > United States > Arizona (0.04)
- Asia > Japan > Honshū > Kansai > Hyogo Prefecture > Kobe (0.04)
- Transportation > Ground > Road (1.00)
- Information Technology (1.00)
- Automobiles & Trucks (0.94)
Explainability of vision-based autonomous driving systems: Review and challenges
Zablocki, Éloi, Ben-Younes, Hédi, Pérez, Patrick, Cord, Matthieu
This survey reviews explainability methods for vision-based self-driving systems. The concept of explainability has several facets and the need for explainability is strong in driving, a safety-critical application. Gathering contributions from several research fields, namely computer vision, deep learning, autonomous driving, explainable AI (X-AI), this survey tackles several points. First, it discusses definitions, context, and motivation for gaining more interpretability and explainability from self-driving systems. Second, major recent state-of-the-art approaches to develop self-driving systems are quickly presented. Third, methods providing explanations to a black-box self-driving system in a post-hoc fashion are comprehensively organized and detailed. Fourth, approaches from the literature that aim at building more interpretable self-driving systems by design are presented and discussed in detail. Finally, remaining open-challenges and potential future research directions are identified and examined.
- North America > United States (0.14)
- North America > Canada > Quebec > Montreal (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- (2 more...)
- Overview (1.00)
- Research Report > Promising Solution (0.34)
- Transportation > Ground > Road (1.00)
- Information Technology > Security & Privacy (1.00)
- Automobiles & Trucks (1.00)
Split-Second 'Phantom' Images Can Fool Tesla's Autopilot
Safety concerns over automated driver-assistance systems like Tesla's usually focus on what the car can't see, like the white side of a truck that one Tesla confused with a bright sky in 2016, leading to the death of a driver. But one group of researchers has been focused on what autonomous driving systems might see that a human driver doesn't--including "phantom" objects and signs that aren't really there, which could wreak havoc on the road. Researchers at Israel's Ben Gurion University of the Negev have spent the last two years experimenting with those "phantom" images to trick semi-autonomous driving systems. They previously revealed that they could use split-second light projections on roads to successfully trick Tesla's driver-assistance systems into automatically stopping without warning when its camera sees spoofed images of road signs or pedestrians. In new research, they've found they can pull off the same trick with just a few frames of a road sign injected on a billboard's video. And they warn that if hackers hijacked an internet-connected billboard to carry out the trick, it could be used to cause traffic jams or even road accidents while leaving little evidence behind.
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks > Manufacturer (1.00)
- Transportation > Passenger (0.88)
- Transportation > Electric Vehicle (0.88)
Gender Bias In the Driving Systems of AI Autonomous Cars - AI Trends
Here's a topic that entails intense controversy, oftentimes sparking loud arguments and heated responses. Do you think that men are better drivers than women, or do you believe that women are better drivers than men? Seems like most of us have an opinion on the matter, one way or another. Stereotypically, men are often characterized as fierce drivers that have a take-no-prisoners attitude, while women supposedly are more forgiving and civil in their driving actions. Depending on how extreme you want to take these tropes, some would say that women shouldn't be allowed on our roadways due to their timidity, while the same could be said that men should not be at the wheel due to their crazed pedal-to-the-metal predilection.
- Automobiles & Trucks (1.00)
- Transportation > Passenger (0.94)
- Transportation > Ground > Road (0.94)
- Information Technology > Robotics & Automation (0.76)
'Hands-free': Automakers race to next level of not-quite-self-driving cars
Autopilot, ProPilot, CoPilot: Automakers have many names for new systems that allow for hands-free driving, but no safety or performance standards to follow as they roll out the most significant changes to vehicle technology in a generation. Spurred by Tesla's success and eager to start profiting from billions spent on autonomous driving research, automakers are accelerating plans to automate routine driving tasks such as cruising on a highway and make them widely available within five years, industry executives said. Most traditional automakers until recently had resisted allowing drivers to take their hands off the steering wheel for extended periods, concerned about product liability claims. Now, hands-free driving systems offer a new and sorely needed source of profit for automakers and suppliers such as Aptiv, especially when this technology is packaged with other extra-cost options. "Consumers are willing to pay extra - sometimes a lot of money - for advanced technology and features that are convenience-oriented rather than strictly focused on safety," IHS principal analyst Jeremy Carlson said.
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks > Manufacturer (1.00)
- Government > Regional Government > North America Government > United States Government (0.51)